2 research outputs found
Learning a Practical SDR-to-HDRTV Up-conversion using New Dataset and Degradation Models
In media industry, the demand of SDR-to-HDRTV up-conversion arises when users
possess HDR-WCG (high dynamic range-wide color gamut) TVs while most
off-the-shelf footage is still in SDR (standard dynamic range). The research
community has started tackling this low-level vision task by learning-based
approaches. When applied to real SDR, yet, current methods tend to produce dim
and desaturated result, making nearly no improvement on viewing experience.
Different from other network-oriented methods, we attribute such deficiency to
training set (HDR-SDR pair). Consequently, we propose new HDRTV dataset (dubbed
HDRTV4K) and new HDR-to-SDR degradation models. Then, it's used to train a
luminance-segmented network (LSN) consisting of a global mapping trunk, and two
Transformer branches on bright and dark luminance range. We also update
assessment criteria by tailored metrics and subjective experiment. Finally,
ablation studies are conducted to prove the effectiveness. Our work is
available at: https://github.com/AndreGuo/HDRTVDM.Comment: Accepted by CVPR202
Redistributing the Precision and Content in 3D-LUT-based Inverse Tone-mapping for HDR/WCG Display
ITM(inverse tone-mapping) converts SDR (standard dynamic range) footage to
HDR/WCG (high dynamic range /wide color gamut) for media production. It happens
not only when remastering legacy SDR footage in front-end content provider, but
also adapting on-theair SDR service on user-end HDR display. The latter
requires more efficiency, thus the pre-calculated LUT (look-up table) has
become a popular solution. Yet, conventional fixed LUT lacks adaptability, so
we learn from research community and combine it with AI. Meanwhile,
higher-bit-depth HDR/WCG requires larger LUT than SDR, so we consult
traditional ITM for an efficiency-performance trade-off: We use 3 smaller LUTs,
each has a non-uniform packing (precision) respectively denser in dark, middle
and bright luma range. In this case, their results will have less error only in
their own range, so we use a contribution map to combine their best parts to
final result. With the guidance of this map, the elements (content) of 3 LUTs
will also be redistributed during training. We conduct ablation studies to
verify method's effectiveness, and subjective and objective experiments to show
its practicability. Code is available at: https://github.com/AndreGuo/ITMLUT.Comment: Accepted in CVMP2023 (the 20th ACM SIGGRAPH European Conference on
Visual Media Production